A Trace-restricted Kronecker-Factored Approximation to Natural Gradient
نویسندگان
چکیده
Second-order optimization methods have the ability to accelerate convergence by modifying gradient through curvature matrix. There been many attempts use second-order for training deep neural networks. In this work, inspired diagonal approximations and factored such as Kronecker-factored Approximate Curvature (KFAC), we propose a new approximation Fisher information matrix (FIM) called Trace-restricted (TKFAC), which can hold certain trace relationship between exact approximate FIM. TKFAC, decompose each block of FIM Kronecker product two smaller matrices scaled coefficient related trace. We theoretically analyze TKFAC's error give an upper bound it. also damping technique TKFAC on convolutional networks maintain superiority during training. Experiments show that our method has better performance compared with several state-of-the-art algorithms some network architectures.
منابع مشابه
Factored 3-Way Restricted Boltzmann Machines For Modeling Natural Images
Deep belief nets have been successful in modeling handwritten characters, but it has proved more difficult to apply them to real images. The problem lies in the restricted Boltzmann machine (RBM) which is used as a module for learning deep belief nets one layer at a time. The Gaussian-Binary RBMs that have been used to model real-valued data are not a good way to model the covariance structure ...
متن کاملScalable trust-region method for deep reinforcement learning using Kronecker-factored approximation
In this work, we propose to apply trust region optimization to deep reinforcement learning using a recently proposed Kronecker-factored approximation to the curvature. We extend the framework of natural policy gradient and propose to optimize both the actor and the critic using Kronecker-factored approximate curvature (K-FAC) with trust region; hence we call our method Actor Critic using Kronec...
متن کاملAn Empirical Analysis of Proximal Policy Optimization with Kronecker-factored Natural Gradients
Deep reinforcement learning methods have shown tremendous success in a large variety tasks, such as Go [Silver et al., 2016], Atari [Mnih et al., 2013], and continuous control [Lillicrap et al., 2015, Schulman et al., 2015]. Policy gradient methods [Williams, 1992] is an important family of methods in model-free reinforcement learning, and the current state-of-the-art policy gradient methods ar...
متن کاملA Kronecker-factored approximate Fisher matrix for convolution layers
Second-order optimization methods such as natural gradient descent have the potential to speed up training of neural networks by correcting for the curvature of the loss function. Unfortunately, the exact natural gradient is impractical to compute for large models, and most approximations either require an expensive iterative procedure or make crude approximations to the curvature. We present K...
متن کاملOptimizing Neural Networks with Kronecker-factored Approximate Curvature
We propose an efficient method for approximating natural gradient descent in neural networks which we call Kronecker-factored Approximate Curvature (K-FAC). K-FAC is based on an efficiently invertible approximation of a neural network’s Fisher information matrix which is neither diagonal nor low-rank, and in some cases is completely non-sparse. It is derived by approximating various large block...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2021
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v35i9.16921